Approximate Planning for Factored POMDPs
نویسندگان
چکیده
We describe an approximate dynamic programming algorithm for partially observable Markov decision processes represented in factored form. Two complementary forms of approximation are used to simplify a piecewise linear and convex value function, where each linear facet of the function is represented compactly by an algebraic decision diagram. ln one form of approximation, the degree of state abstraction is increased by aggregating states with similar values. In the second form of approximation, the value function is simplified by removing linear facets that contribute marginally to value. We derive an error bound that applies to both forms of approximation. Experimental results show that this approach improves the performance of dynamic programming and extends the range of problems it can
منابع مشابه
Approximate Planning for Factored POMDPs using Belief State Simpli cation
We are interested in the problem of planning for factored POMDPs. Building on the recent results of Kearns, Mansour and Ng, we provide a planning algorithm for fac-tored POMDPs that exploits the accuracy-eeciency tradeoo in the belief state simplii-cation introduced by Boyen and Koller.
متن کاملApproximate Solutions for Factored Dec-POMDPs with Many Agents1
Dec-POMDPs are a powerful framework for planning in multiagent systems, but are provably intractable to solve. This paper proposes a factored forward-sweep policy computation method that tackles the stages of the problem one by one, exploiting weakly coupled structure at each of these stages. An empirical evaluation shows that the loss in solution quality due to these approximations is small an...
متن کاملApproximate Solutions for Factored Dec-POMDPs with Many Agents — Extended Abstract1
Dec-POMDPs are a powerful framework for planning in multiagent systems, but are provably intractable to solve. This paper proposes a factored forward-sweep policy computation method that tackles the stages of the problem one by one, exploiting weakly coupled structure at each of these stages. An empirical evaluation shows that the loss in solution quality due to these approximations is small an...
متن کاملApproximate Planning for Factored POMDPs using Belief State Simplification
We are interested in the problem of plan ning for factored POMOPs. Building on the recent results of Kearns, Mansour and Ng, we provide a planning algorithm for fac tored POMOPs that exploits the accuracy efficiency tradeoff in the belief state sim plifi cation introduced by Boyen and Koller.
متن کاملAn Approach to State Aggregation for POMDPs
A partially observable Markov decision process (POMDP) provides an elegant model for problems of planning under uncertainty. Solving POMDPs is very computationally challenging, however, and improving the scalability of POMDP algorithms is an important research problem. One way to reduce the computational complexity of planning using POMDPs is by using state aggregation to reduce the (effective)...
متن کاملApproximate solutions for factored Dec-POMDPs with many agents
Dec-POMDPs are a powerful framework for planning in multiagent systems, but are provably intractable to solve. Despite recent work on scaling to more agents by exploiting weak couplings in factored models, scalability for unrestricted subclasses remains limited. This paper proposes a factored forward-sweep policy computation method that tackles the stages of the problem one by one, exploiting w...
متن کامل